Skip to content

Updated Megatron version #85

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Draft
wants to merge 432 commits into
base: nvidia_main
Choose a base branch
from

Conversation

jlamypoirier
Copy link
Collaborator

No description provided.

shjwudp and others added 30 commits January 17, 2024 09:16
Refactor DistributedOptimizer for MoE model support

See merge request ADLR/megatron-lm!986
Signed-off-by: Selvaraj Anandaraj <selvaraja@login-eos01.eos.clusters.nvidia.com>
Signed-off-by: jiemingz <jiemingz@nvidia.com>
Signed-off-by: jiemingz <jiemingz@nvidia.com>
Signed-off-by: jiemingz <jiemingz@nvidia.com>
Signed-off-by: Selvaraj Anandaraj <selvaraja@login-eos01.eos.clusters.nvidia.com>
add is_first_microbatch for TE

See merge request ADLR/megatron-lm!1033
Need a switch at NeMo level to enable Atomic GEMM

See merge request ADLR/megatron-lm!1017
Add distributed checkpoint support to non-TE based models

See merge request ADLR/megatron-lm!1005
Signed-off-by: Selvaraj Anandaraj <selvaraja@login-eos01.eos.clusters.nvidia.com>
Signed-off-by: Selvaraj Anandaraj <selvaraja@login-eos01.eos.clusters.nvidia.com>
Support for activation offloading to CPU in M-LM

See merge request ADLR/megatron-lm!1016
add rope and swiglu fusion

See merge request ADLR/megatron-lm!946
Add jit_fuser to switch between torch.jit.script and torch.compile

See merge request ADLR/megatron-lm!1036
Run black on megatron/optimizer

See merge request ADLR/megatron-lm!1050
deepakn94 and others added 30 commits February 25, 2024 23:00
Unify resume and correctness functional tests

See merge request ADLR/megatron-lm!1070
Mcore mock multimodal dataset

See merge request ADLR/megatron-lm!1147
…ommunication

Compute norm once per batch (instead of once per microbatch) and once per bucket (instead of once per param)
Fix NaN checking in grads: should be performed before data-parallel all-reduce

See merge request ADLR/megatron-lm!989
Move to Draco OCI

See merge request ADLR/megatron-lm!1137
Print number of transformer and embedding parameters separately

See merge request ADLR/megatron-lm!1159
Mcore LLaVA model

See merge request ADLR/megatron-lm!1151
[OMNIML-614] AMMO ptq + TensorRT-LLM export examples for megatron-lm

See merge request ADLR/megatron-lm!1013
Make throughput and memory footprint formulae compatible with arbitrary ffn_hidden_size

See merge request ADLR/megatron-lm!1169
Experimental Yaml configs

See merge request ADLR/megatron-lm!1134
This reverts commit fe1f23c.
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.